Expectation Maximization and Posterior Constraints

نویسندگان

  • João Graça
  • Kuzman Ganchev
  • Ben Taskar
چکیده

The expectation maximization (EM) algorithm is a widely used maximum likelihood estimation procedure for statistical models when the values of some of the variables in the model are not observed. Very often, however, our aim is primarily to find a model that assigns values to the latent variables that have intended meaning for our data and maximizing expected likelihood only sometimes accomplishes this. Unfortunately, it is typically difficult to add even simple a-priori information about latent variables in graphical models without making the models overly complex or intractable. In this paper, we present an efficient, principled way to inject rich constraints on the posteriors of latent variables into the EM algorithm. Our method can be used to learn tractable graphical models that satisfy additional, otherwise intractable constraints. Focusing on clustering and the alignment problem for statistical machine translation, we show that simple, intuitive posterior constraints can greatly improve the performance over standard baselines and be competitive with more complex, intractable models.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A Framework for Tuning Posterior Entropy in Unsupervised Learning

We present a general framework for unsupervised and semi-supervised learning containing a graded spectrum of Expectation Maximization (EM) algorithms. We call our framework Unified Expectation Maximization (UEM.) UEM allows us to tune the entropy of the inferred posterior distribution during the E-step to impact the quality of learning. Furthermore, UEM covers existing algorithms like standard ...

متن کامل

Homeostatic plasticity in Bayesian spiking networks as Expectation Maximization with posterior constraints

Recent spiking network models of Bayesian inference and unsupervised learning frequently assume either inputs to arrive in a special format or employ complex computations in neuronal activation functions and synaptic plasticity rules. Here we show in a rigorous mathematical treatment how homeostatic processes, which have previously received little attention in this context, can overcome common ...

متن کامل

An Efficient Posterior Regularized Latent Variable Model for Interactive Sound Source Separation

In applications such as audio denoising, music transcription, music remixing, and audiobased forensics, it is desirable to decompose a single-channel recording into its respective sources. One of the current most effective class of methods to do so is based on nonnegative matrix factorization and related latent variable models. Such techniques, however, typically perform poorly when no isolated...

متن کامل

Semi-blind Channel Identification and Equalization in Ofdm: an Expectation-maximization Approach

This work proposes an expectation-maximization approach to channel identification and equalization in OFDM. The algorithm exploits the natural constraints imposed by the channel (sparsity, maximum delay spread, and a priori statistical information) and those imposed by the transmitter (pilots, cyclic prefix, and the finite alphabet constraint). These constraints are used to reduce the number of...

متن کامل

A Representation Approach for Relative Entropy Minimization with Expectation Constraints

We consider the general problem of relative entropy minimization and entropy maximization subject to expectation constraints. We show that the solutions can be represented as members of an exponential family subject to weaker conditions than previously shown, and the representation can be simplified further if an appropriate conjugate prior density is used. As a result, the solutions can be fou...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2007